Explaining recommendations generated by MDPs
نویسندگان
چکیده
There has been little work in explaining recommendations generated by Markov Decision Processes (MDPs). We analyze the difculty of explaining policies computed automatically and identify a set of templates that can be used to generate explanations automatically at run-time. These templates are domain-independent and can be used in any application of an MDP. We show that no additional e ort is required from the MDP designer for producing such explanations. We use the problem of advising undergraduate students in their course selection to explain the recommendation for selecting speci c courses to students. We also propose an extension to leverage domain-speci c constructs using ontologies so that explanations can be made more user-friendly.
منابع مشابه
Minimal Sufficient Explanations for Factored Markov Decision Processes
Explaining policies of Markov Decision Processes (MDPs) is complicated due to their probabilistic and sequential nature. We present a technique to explain policies for factored MDP by populating a set of domain-independent templates. We also present a mechanism to determine a minimal set of templates that, viewed together, completely justify the policy. Our explanations can be generated automat...
متن کاملPolicy Explanation and Model Refinement in Decision-Theoretic Planning
Decision-theoretic systems, such as Markov Decision Processes (MDPs), are used for sequential decision-making under uncertainty. MDPs provide a generic framework that can be applied in various domains to compute optimal policies. This thesis presents techniques that offer explanations of optimal policies for MDPs and then refine decision theoretic models (Bayesian networks and MDPs) based on fe...
متن کاملAutomated Explanations for MDP Policies
Explaining policies of Markov Decision Processes (MDPs) is complicated due to their probabilistic and sequential nature. We present a technique to explain policies for factored MDP by populating a set of domain-independent templates. We also present a mechanism to determine a minimal set of templates that, viewed together, completely justify the policy. We demonstrate our technique using the pr...
متن کاملAbstraction and Approximate Decision-Theoretic Planning
ion and Approximate Decision Theoretic Planning Richard Dearden and Craig Boutiliery Department of Computer Science University of British Columbia Vancouver, British Columbia CANADA, V6T 1Z4 email: dearden,[email protected] Abstract Markov decision processes (MDPs) have recently been proposed as useful conceptual models for understanding decision-theoretic planning. However, the utility of the as...
متن کاملAn Empirical Analysis of Off-policy Learning in Discrete MDPs
Abstract Off-policy evaluation is the problem of evaluating a decision-making policy using data collected under a different behaviour policy. While several methods are available for addressing off-policy evaluation, little work has been done on identifying the best methods. In this paper, we conduct an in-depth comparative study of several off-policy evaluation methods in non-bandit, finite-hor...
متن کامل